Spectral Convolution Networks
نویسندگان
چکیده
Previous research has shown that computation of convolution in the frequency domain provides a significant speedup versus traditional convolution network implementations. However, this performance increase comes at the expense of repeatedly computing the transform and its inverse in order to apply other network operations such as activation, pooling, and dropout. We show, mathematically, how convolution and activation can both be implemented in the frequency domain using either the Fourier or Laplace transformation. The main contributions are a description of spectral activation under the Fourier transform and a further description of an efficient algorithm for computing both convolution and activation under the Laplace transform. By computing both the convolution and activation functions in the frequency domain, we can reduce the number of transforms required, as well as reducing overall complexity. Our description of a spectral activation function, together with existing spectral analogs of other network functions may then be used to compose a fully spectral implementation of a convolution network. 1 Motivation Convolution networks are used for machine learning problems such as image classification, natural language processing, and recommendation systems [4, 3, 8]. They are represented as a graph of operators which are typically sequentially applied to some input image, eventually yielding a classification for that input. Convolution is an expensive operation which is replicated repeatedly within a single network. Computation of convolution in the frequency domain under the Fourier transform has been shown to provide a significant speedup versus traditional convolution network implementations [9, 4]. However, typically activation is run following convolution in practise, and previous researchers have been unable to find a spectral implementation for both. In this paper we describe spectral representations of the max activation function a(x) paying particular attention to computational complexity. Specifically, a(x) = max(0, x). The abbreviation ReLU is used throughout to designate the part of the network which computes this function.
منابع مشابه
Numerical resolution of large deflections in cantilever beams by Bernstein spectral method and a convolution quadrature.
The mathematical modeling of the large deflections for the cantilever beams leads to a nonlinear differential equation with the mixed boundary conditions. Different numerical methods have been implemented by various authors for such problems. In this paper, two novel numerical techniques are investigated for the numerical simulation of the problem. The first is based on a spectral method utiliz...
متن کاملSome Results on Convex Spectral Functions: I
In this paper, we give a fundamental convexity preserving for spectral functions. Indeed, we investigate infimal convolution, Moreau envelope and proximal average for convex spectral functions, and show that this properties are inherited from the properties of its corresponding convex function. This results have many applications in Applied Mathematics such as semi-definite programmings and eng...
متن کاملWavelet Convolutional Neural Networks for Texture Classification
Texture classification is an important and challenging problem in many image processing applications. While convolutional neural networks (CNNs) achieved significant successes for image classification, texture classification remains a difficult problem since textures usually do not contain enough information regarding the shape of object. In image processing, texture classification has been tra...
متن کاملBipolar spectral associative memories
Nonlinear spectral associative memories are proposed as quantized frequency domain formulations of nonlinear, recurrent associative memories in which volatile network attractors are instantiated by attractor waves. In contrast to conventional associative memories, attractors encoded in the frequency domain by convolution may be viewed as volatile online inputs, rather than nonvolatile, off-line...
متن کاملInferring Phonemic Classes from CNN Activation Maps Using Clustering Techniques
Today’s state-of-art in speech recognition involves deep neural networks (DNN). These last years, a certain research effort has been invested in characterizing the feature representations learned by DNNs. In this paper, we focus on convolutional neural networks (CNN) trained for phoneme recognition in French. We report clustering experiments performed on activation maps extracted from the diffe...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- CoRR
دوره abs/1611.05378 شماره
صفحات -
تاریخ انتشار 2016